# Whole Word Masking

Rbt4 H312
Apache-2.0
MiniRBT is a Chinese small pre-trained model developed based on knowledge distillation technology, optimized for training efficiency using Whole Word Masking.
Large Language Model Transformers Chinese
R
hfl
34
5
Minirbt H288
Apache-2.0
MiniRBT is a Chinese small pretrained model developed based on knowledge distillation technology, optimized for training efficiency using Whole Word Masking.
Large Language Model Transformers Chinese
M
hfl
405
8
Minirbt H256
Apache-2.0
MiniRBT is a small Chinese pre-trained model based on knowledge distillation technology, combined with whole word masking, suitable for various Chinese natural language processing tasks.
Large Language Model Transformers Chinese
M
hfl
225
7
Erlangshen DeBERTa V2 320M Chinese
Apache-2.0
Chinese pre-trained language model based on DeBERTa-v2 architecture with 320 million parameters, excelling in natural language understanding tasks
Large Language Model Transformers Chinese
E
IDEA-CCNL
186
13
Erlangshen DeBERTa V2 97M Chinese
Apache-2.0
A Chinese DeBERTa-v2 base model specialized in natural language understanding tasks, employing Whole Word Masking with 97 million parameters.
Large Language Model Transformers Chinese
E
IDEA-CCNL
178
11
Bert Base Swedish Cased Ner
Swedish BERT base model released by the National Library of Sweden/KBLab, trained on multi-source texts
Large Language Model Other
B
KB
20.77k
8
Bert Large Uncased Whole Word Masking Squad Int8 0001
BERT-large English Q&A model pre-trained with whole word masking and fine-tuned on SQuAD v1.1, quantized to INT8 precision
Question Answering System Transformers
B
dkurt
23
0
Chinese Bert Wwm
Apache-2.0
A Chinese pre-trained BERT model using whole word masking strategy, designed to accelerate Chinese natural language processing research.
Large Language Model Chinese
C
hfl
28.52k
79
Bert Base Spanish Wwm Cased
BETO is a BERT model trained on a large Spanish corpus, available in both cased and uncased versions, suitable for various Spanish NLP tasks.
Large Language Model Spanish
B
dccuchile
61.43k
63
Bert Base Spanish Wwm Uncased
BETO is a BERT model trained on a large Spanish corpus, supporting both cased and uncased versions, suitable for various Spanish NLP tasks.
Large Language Model Spanish
B
dccuchile
231.26k
65
Bert Base Japanese Basic Char V2
This is a Japanese BERT pre-trained model based on character-level tokenization and whole word masking techniques, requiring no dependency on `fugashi` or `unidic_lite` toolkits.
Large Language Model Transformers Japanese
B
hiroshi-matsuda-rit
14
0
Umberto Commoncrawl Cased V1
Italian language model based on the Roberta architecture, trained with SentencePiece tokenization and Whole Word Masking
Large Language Model Transformers Other
U
Musixmatch
13.19k
15
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase